5 research outputs found
Simultaneous lesion and neuroanatomy segmentation in Multiple Sclerosis using deep neural networks
Segmentation of both white matter lesions and deep grey matter structures is
an important task in the quantification of magnetic resonance imaging in
multiple sclerosis. Typically these tasks are performed separately: in this
paper we present a single segmentation solution based on convolutional neural
networks (CNNs) for providing fast, reliable segmentations of multimodal
magnetic resonance images into lesion classes and normal-appearing grey- and
white-matter structures. We show substantial, statistically significant
improvements in both Dice coefficient and in lesion-wise specificity and
sensitivity, compared to previous approaches, and agreement with individual
human raters in the range of human inter-rater variability. The method is
trained on data gathered from a single centre: nonetheless, it performs well on
data from centres, scanners and field-strengths not represented in the training
dataset. A retrospective study found that the classifier successfully
identified lesions missed by the human raters.
Lesion labels were provided by human raters, while weak labels for other
brain structures (including CSF, cortical grey matter, cortical white matter,
cerebellum, amygdala, hippocampus, subcortical GM structures and choroid
plexus) were provided by Freesurfer 5.3. The segmentations of these structures
compared well, not only with Freesurfer 5.3, but also with FSL-First and
Freesurfer 6.0
Simultaneous lesion and brain segmentation in multiple sclerosis using deep neural networks.
Segmentation of white matter lesions and deep grey matter structures is an important task in the quantification of magnetic resonance imaging in multiple sclerosis. In this paper we explore segmentation solutions based on convolutional neural networks (CNNs) for providing fast, reliable segmentations of lesions and grey-matter structures in multi-modal MR imaging, and the performance of these methods when applied to out-of-centre data. We trained two state-of-the-art fully convolutional CNN architectures on the 2016 MSSEG training dataset, which was annotated by seven independent human raters: a reference implementation of a 3D Unet, and a more recently proposed 3D-to-2D architecture (DeepSCAN). We then retrained those methods on a larger dataset from a single centre, with and without labels for other brain structures. We quantified changes in performance owing to dataset shift, and changes in performance by adding the additional brain-structure labels. We also compared performance with freely available reference methods. Both fully-convolutional CNN methods substantially outperform other approaches in the literature when trained and evaluated in cross-validation on the MSSEG dataset, showing agreement with human raters in the range of human inter-rater variability. Both architectures showed drops in performance when trained on single-centre data and tested on the MSSEG dataset. When trained with the addition of weak anatomical labels derived from Freesurfer, the performance of the 3D Unet degraded, while the performance of the DeepSCAN net improved. Overall, the DeepSCAN network predicting both lesion and anatomical labels was the best-performing network examined